We broke AI guardrails down to six categories.
We curated datasets and models that demonstrate the state of AI safety using LLMs and other open source models.
Developer | Model | Latency | Metric |
---|---|---|---|
Guardrails AI | Toxic Language | 0.0086 ms | 0.7182 |
Natural Language Content Safety | 0.1063 ms | 0.5955 | |
Microsoft | Azure Content Safety | 0.0632 ms | 0.5125 |
Developer | Samples |
---|---|
toxic | 6090 |
obscene | 3691 |
insult | 3427 |
identity_hate | 712 |
severe_toxic | 367 |
threat | 211 |